Optimal learning via local entropies and sample compression

نویسنده

  • Zhivotovskiy Nikita
چکیده

Under margin assumptions, we prove several risk bounds, represented via the distribution depen-dent local entropies of the classes or the sizes of specific sample compression schemes. In somecases, our guarantees are optimal up to constant factors for families of classes. We discuss limi-tations of our approach and give several applications. In particular, we provide a new tight PACbound for the hard-margin SVM, an extended analysis of certain empirical risk minimizers underlog-concave distributions, a new variant of an online to batch conversion, and distribution depen-dent localized bounds in the aggregation framework. As a part of our results, we give a new upperbound for the uniform deviations under Bernstein assumptions, which may be of independent in-terest. The proofs for the sample compression schemes are based on the moment method combinedwith the analysis of voting algorithms.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Improving Recognition via Reconstruction and Other Constraints

Learning a many-parameter model is generally an under-constrained problem that requires additional regularization. We study several information theoretic constraints and show that reconstruction constraints achieve improved performance. In addition to studying entropy and BCM constraints on the hidden units of a feed-forward architecture, we derive a new constraint that is based on the sum of t...

متن کامل

Deblocking Joint Photographic Experts Group Compressed Images via Self-learning Sparse Representation

JPEG is one of the most widely used image compression method, but it causes annoying blocking artifacts at low bit-rates. Sparse representation is an efficient technique which can solve many inverse problems in image processing applications such as denoising and deblocking. In this paper, a post-processing method is proposed for reducing JPEG blocking effects via sparse representation. In this ...

متن کامل

A new universal coding scheme for the binary memoryless source

A coding scheme for the binary memoryless source is presented. The scheme is an extension of a scheme discovered by Lynch and Davisson and developed by Schalkwijk, and is based upon the Schalkwijk algorithm for the Lynch-Davisson (LD) code. The new coding scheme is shown to be asymptotically optimal as the block length approaches infinity, regardless of the source statistics. Although the LD sc...

متن کامل

On the Statistical Estimation of Rényi Entropies

Estimating entropies is important in many fields including statistical physics, machine learning and statistics. While the Shannon logarithmic entropy is the most fundamental, other Rényi entropies are also of importance. In this paper, we derive a bias corrected estimator for a subset of Rényi entropies. The advantage of the estimator is demonstrated via theoretical and experimental considerat...

متن کامل

On statistical learning via the lens of compression

This work continues the study of the relationship between sample compression schemes and statistical learning, which has been mostly investigated within the framework of binary classification. The central theme of this work is establishing equivalences between learnability and compressibility, and utilizing these equivalences in the study of statistical learning theory. We begin with the settin...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2017